Preliminary Investigation of Distributed Shared Memory System on a Cluster of High Performance Clusters

نویسندگان

  • Takeshi Nanri
  • Masaaki Shimasaki
  • Yoshitaka Watanabe
  • Hiyoyuki Sato
چکیده

This paper introduces design and basic performance of the DSM(distributed shared memory) system on a cluster of clusters. Networking devices such as Myrinet have improved the performance of cluster systems significantly. In addition to that, such kind of network devices introduced a new hierarchical architecture; a multi-cluster, a cluster of high-performance clusters. To ease the difficulty of programming with message passing, which is the conventional programming paradigm on cluster systems, many DSM (distributed shared memory) systems have been developed in recent years. However, there have been no DSM systems developed on multi-clusters. The DSM system consists of a runtime system to support basic functions for accessing virtual shared memory built on such environment. The functions are allocation of global data, read and write accesses to global data, synchronization of the whole system, and mutual exclusion. The authors have evaluated the performance of the runtime system, built on a SMP cluster, COMPaS, at RWCP(Real World Computing Partnership) in Tsukuba, Japan. The result shows that a read access to remote memory on the same cluster costs about 0.2msec, while a read access to remote memory on other cluster costs about 1.3msec. The execution time of LU decomposition on a multi-cluster consisting two clusters of three PCs is about 2.8times faster than the time on one PC.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Hierarchical Shared Memory Cluster Architecture with Load Balancing and Fault Tolerance

Recently a great deal of attention has been paid to the design of hierarchical shared memory cluster system. Cluster computing has made hierarchical computing systems increasingly common as target environment for large-scale scientific computations. This paper proposes hierarchical shared memory cluster architecture with load balancing and fault tolerance. Hierarchies of shared memory and cache...

متن کامل

Strings: A High-Performance Distributed Shared Memory for Symmetrical Multiprocessor Clusters

This paper introduces Strings, a high performance distributed shared memory system designed for clusters of symmetrical multiprocessors (SMPs). The distinguishing feature of this system is the use of a fully multi-threaded runtime system, written using POSIX threads. Strings also allows multiple application threads to be run on each node in a cluster. Since most modern UNIX systems can multiple...

متن کامل

A Distributed Shared Memory Cluster Architecture With Dynamic Load Balancing

This paper proposes a distributed shared memory cluster architecture with load balancing. The architecture is based on dynamic task scheduling approach for distribution and assignment. It enhances the performance of communication across clusters for data access. The proposed dynamic load balancing model uses the concept of work stealing, which intelligently balances the load among different nod...

متن کامل

Design of OpenMP Compiler for an SMP Cluster

In this paper, we present a design of OpenMP compiler for an SMP cluster. Although clusters of SMPs are expected to be one of the cost-e ective parallel computing platforms, both of inter and intra node parallelism must be exploited to achieve high performance. These two levels of structure complicate parallel programming. The OpenMP is an emerging standard for parallel programming on shared-me...

متن کامل

An OpenMP Programming Toolkit for Hybrid CPU/GPU Clusters Based on Software Unified Memory

Recently, hybrid CPU/GPU cluster has drawn much attention from the researchers of high performance computing because of amazing energy efficiency and adaptable resource exploitation. However, the programming of hybrid CPU/GPU clusters is very complex because it requires users to learn new programming interfaces such as CUDA and OpenCL, and combine them with MPI and OpenMP. To address this probl...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2000